Self-Enhancing Multi-filter Sequence-to-Sequence Model

نویسندگان

چکیده

Representation learning is important for solving sequence-to-sequence problems in natural language processing. transforms raw data into vector-form representations while preserving their features. However, with significantly different features leads to heterogeneity representations, which may increase the difficulty of convergence. We design a multi-filter encoder-decoder model resolve problem tasks. The divides latent space subspaces using clustering algorithm and trains set decoders (filters) each decoder only concentrates on from its corresponding subspace. As main contribution, we self-enhancing mechanism that uses reinforcement optimize without additional training data. run semantic parsing machine translation experiments indicate proposed can outperform most benchmarks by at least 5%. also empirically show improve performance over 10% provide evidence demonstrate positive correlation between model's clustering.

منابع مشابه

A Mathematical Model to Optimize Allocation Sequence in Dispatching Problem

Truck-Shovel fleet, as the most common transportation system in open-pit mines, has a significant part of mining costs, for which optimal management can lead to substantial cost reductions. Among the available dispatch mathematical models, the multi-stage approach is well suited for allocating trucks to respected shovels in a dynamic dispatching program. However, with this kind of modeling sequ...

متن کامل

Multi-task Sequence to Sequence Learning

Sequence to sequence learning has recently emerged as a new paradigm in supervised learning. To date, most of its applications focused on only one task and not much work explored this framework for multiple tasks. This paper examines three settings to multi-task sequence to sequence learning: (a) the one-to-many setting – where the encoder is shared between several tasks such as machine transla...

متن کامل

Sequence to Sequence Model for Video Captioning

Automatically generating video captions with natural language remains a challenge for both the field of nature language processing and computer vision. Recurrent Neural Networks (RNNs), which models sequence dynamics, has proved to be effective in visual interpretation. Based on a recent sequence to sequence model for video captioning, which is designed to learn the temporal structure of the se...

متن کامل

Gamma-filter self-organising neural networks for unsupervised sequence processing

Adding g-filters to self-organising neural networks for unsupervised sequence processing is proposed. The proposed g-context model is applied to self-organising maps and neural gas networks. The g-context model is a generalisation that includes as a particular example the previously published merge-context model. The results show that the g-context model outperforms the merge-context model in t...

متن کامل

Multi-Dialect Speech Recognition With A Single Sequence-To-Sequence Model

Sequence-to-sequence models provide a simple and elegant solution for building speech recognition systems by folding separate components of a typical system, namely acoustic (AM), pronunciation (PM) and language (LM) models into a single neural network. In this work, we look at one such sequence-to-sequence model, namely listen, attend and spell (LAS) [1], and explore the possibility of trainin...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Procedia Computer Science

سال: 2022

ISSN: ['1877-0509']

DOI: https://doi.org/10.1016/j.procs.2022.12.056